Network-based dimensionality reduction of high-dimensional, low-sample-size datasets

نویسندگان

چکیده

In the field of data science, there are a variety datasets that suffer from high-dimensional, low-sample-size (HDLSS) problem; however, only few dimensionality reduction methods exist applicable to address this type problem, and is no nonparametric solution date. The purpose work develop novel network-based (nonparametric) analysis (NDA) method, can be effectively applied HDLSS data. First, with NDA correlation graph variables specified. With modularity-based community detection set modules Then, linear combination weighted by their eigenvector centralities (EVCs), defined as LVs, determined. optional phase variable selection, low EVCs communality ignored. LVs indicators belonging specified using method. publicly available databases compared principal factoring (PFA) methods. results show it outperforms existing in terms interpretability. addition, application easier, since need specify number latent due its nature. • A method proposed perform reduction. finds (LVs) indicators. provides feature ignoring common communalities. both LVs. tested component on databases.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An Efficient Dimensionality Reduction Approach for Small-sample Size and High-dimensional Data Modeling

As for massive multidimensional data are being generated in a wide range of emerging applications, this paper introduces two new methods of dimension reduction to conduct small-sample size and high-dimensional data processing and modeling. Through combining the support vector machine (SVM) and recursive feature elimination (RFE), SVM-RFE algorithm is proposed to select features, and further, ad...

متن کامل

High Dimensional Correspondences from Low Dimensional Manifolds - An Empirical Comparison of Graph-Based Dimensionality Reduction Algorithms

Linear methods yield data points xn = r1φn,1 +r ∗ 2φn,2 + . . .+ rMφn,M = R φn, which are d-dimensional linear combinations of the original D-dimensional data points φn. . Principal Component Analysis (PCA, Jolliffe (2002)): Preserves the global covariance structure by decomposition of the covariance matrix Σφ,φ = RS R. . Metric Multidimensional Scaling (MDS, Cox & Cox (1994)): Preserves inner ...

متن کامل

Efficient Dimensionality Reduction for High-Dimensional Network Estimation

We propose module graphical lasso (MGL), an aggressive dimensionality reduction and network estimation technique for a highdimensional Gaussian graphical model (GGM). MGL achieves scalability, interpretability and robustness by exploiting the modularity property of many real-world networks. Variables are organized into tightly coupled modules and a graph structure is estimated to determine the ...

متن کامل

The Novel WFCM Algorithm for Dimensionality Reduction of High Dimensional Datasets

Abstract—Dimensionality reduction studies techniques that successfully reduce data dimensionality for proficient data processing assignments such as pattern recognition (PR) machine learning (ML), text retrieval, and data mining (DM). From last many years Broad research into dimensionality reduction is being carried out and presently it’s also in demand for additional growing due to imperative ...

متن کامل

Cox model in high dimensional and low sample size settings

The evolution of cancer is more certainly linked to a complex interplay of genes rather than a single gene activity. Multivariate analysis, which can exploit the correlated pattern of gene expression display by genes behaving jointly, such as genes performing the same functions or genes operating along the same pathway, can become a very useful diagnostic tool to determine molecular predictor o...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Knowledge Based Systems

سال: 2022

ISSN: ['1872-7409', '0950-7051']

DOI: https://doi.org/10.1016/j.knosys.2022.109180